LLM 0.26 introduces tool support, allowing LLMs to access and utilize Python functions as tools. The article details how to install, configure, and use these tools with various LLMs like OpenAI, Anthropic, Gemini, and Ollama models, including examples with plugins and ad-hoc functions. It also discusses the implications for building 'agents' and future development plans.
This tutorial demonstrates how to integrate Google’s Gemini 2.0 with an in-process Model Context Protocol (MCP) server using FastMCP, creating tools for weather information and integrating them into Gemini's function calling workflow.
Learn how to build an open LLM app using Hermes 2 Pro, a powerful LLM based on Meta's Llama 3 architecture. This tutorial explains how to deploy Hermes 2 Pro locally, create a function to track flight status using FlightAware API, and integrate it with the LLM.
This article describes the process of implementing function-calling in an AI system, specifically using the Mistral AI platform. The example showcases the development of an assistant that can manage a home automation system through natural language interactions with the user, including the use of available functions, function logic, and the integration of these functions into the AI system.
A tutorial showing you how how to bring real-time data to LLMs through function calling, using OpenAI's latest LLM GTP-4o.